Sampling Methods for Efficient Training of Graph Convolutional Networks: A Survey

نویسندگان

چکیده

Graph convolutional networks (GCNs) have received significant attention from various research fields due to the excellent performance in learning graph representations. Although GCN performs well compared with other methods, it still faces challenges. Training a model for large-scale graphs conventional way requires high computation and storage costs. Therefore, motivated by an urgent need terms of efficiency scalability training GCN, sampling methods been proposed achieved effect. In this paper, we categorize based on mechanisms provide comprehensive survey efficient GCN. To highlight characteristics differences present detailed comparison within each category further give overall comparative analysis all categories. Finally, discuss some challenges future directions methods.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Stochastic Training of Graph Convolutional Networks

Graph convolutional networks (GCNs) are powerful deep neural networks for graph-structured data. However, GCN computes nodes’ representation recursively from their neighbors, making the receptive field size grow exponentially with the number of layers. Previous attempts on reducing the receptive field size by subsampling neighbors do not have any convergence guarantee, and their receptive field...

متن کامل

FastGCN: Fast Learning with Graph Convolutional Networks via Importance Sampling

The graph convolutional networks (GCN) recently proposed by Kipf and Welling are an effective graph model for semi-supervised learning. This model, however, was originally designed to be learned with the presence of both training and test data. Moreover, the recursive neighborhood expansion across layers poses time and memory challenges for training with large, dense graphs. To relax the requir...

متن کامل

Stochastic Training of Graph Convolutional Networks with Variance Reduction

Graph convolutional networks (GCNs) are powerful deep neural networks for graph-structured data. However, GCN computes the representation of a node recursively from its neighbors, making the receptive field size grow exponentially with the number of layers. Previous attempts on reducing the receptive field size by subsampling neighbors do not have a convergence guarantee, and their receptive fi...

متن کامل

Dynamic Graph Convolutional Networks

Many different classification tasks need to manage structured data, which are usually modeled as graphs. Moreover, these graphs can be dynamic, meaning that the vertices/edges of each graph may change during time. Our goal is to jointly exploit structured data and temporal information through the use of a neural network model. To the best of our knowledge, this task has not been addressed using...

متن کامل

Graph Convolutional Networks

We present a scalable approach for semi-supervised learning on graph-structured data that is based on an efficient variant of convolutional neural networks which operate directly on graphs. We motivate the choice of our convolutional architecture via a localized first-order approximation of spectral graph convolutions. Our model scales linearly in the number of graph edges and learns hidden lay...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE/CAA Journal of Automatica Sinica

سال: 2022

ISSN: ['2329-9274', '2329-9266']

DOI: https://doi.org/10.1109/jas.2021.1004311